extracting hard confident example
Appendix for " CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation " Y exiong Lin 1 Y u Y ao
We denote observed variables with gray color and latent variables with white color. Firstly, we introduce the concept of an uncontrolled style factor . Why do confident examples encourage content-style isolation? Calculate the loss using Eq. 1 and update networks; Output: The inference networks and classifier heads q It's essential to understand that although data augmentation cannot control all style factors, it still offers the benefit of "partial isolation". This approach, therefore, ensures that styles changes don't affect the derived content representation Calculate the loss using Eq. 2 and update networks; Output: The inference networks and classifier heads q Finally, confident and unlabeled examples are used to train the models based on the MixMatch algorithm.
- North America > United States (0.05)
- Asia > China > Hong Kong (0.04)
CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation
Label noise widely exists in large-scale image datasets. To mitigate the side effects of label noise, state-of-the-art methods focus on selecting confident examples by leveraging semi-supervised learning. Existing research shows that the ability to extract hard confident examples, which are close to the decision boundary, significantly influences the generalization ability of the learned classifier.In this paper, we find that a key reason for some hard examples being close to the decision boundary is due to the entanglement of style factors with content factors. The hard examples become more discriminative when we focus solely on content factors, such as semantic information, while ignoring style factors. Nonetheless, given only noisy data, content factors are not directly observed and have to be inferred.To tackle the problem of inferring content factors for classification when learning with noisy labels, our objective is to ensure that the content factors of all examples in the same underlying clean class remain unchanged as their style information changes.To achieve this, we utilize different data augmentation techniques to alter the styles while regularizing content factors based on some confident examples. By training existing methods with our inferred content factors, CS-Isolate proves their effectiveness in learning hard examples on benchmark datasets. The implementation is available at https://github.com/tmllab/2023
Appendix for " CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation " Y exiong Lin 1 Y u Y ao
We denote observed variables with gray color and latent variables with white color. Firstly, we introduce the concept of an uncontrolled style factor . Why do confident examples encourage content-style isolation? Calculate the loss using Eq. 1 and update networks; Output: The inference networks and classifier heads q It's essential to understand that although data augmentation cannot control all style factors, it still offers the benefit of "partial isolation". This approach, therefore, ensures that styles changes don't affect the derived content representation Calculate the loss using Eq. 2 and update networks; Output: The inference networks and classifier heads q Finally, confident and unlabeled examples are used to train the models based on the MixMatch algorithm.
- North America > United States (0.05)
- Asia > China > Hong Kong (0.04)
CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation
Label noise widely exists in large-scale image datasets. To mitigate the side effects of label noise, state-of-the-art methods focus on selecting confident examples by leveraging semi-supervised learning. Existing research shows that the ability to extract hard confident examples, which are close to the decision boundary, significantly influences the generalization ability of the learned classifier.In this paper, we find that a key reason for some hard examples being close to the decision boundary is due to the entanglement of style factors with content factors. The hard examples become more discriminative when we focus solely on content factors, such as semantic information, while ignoring style factors. Nonetheless, given only noisy data, content factors are not directly observed and have to be inferred.To tackle the problem of inferring content factors for classification when learning with noisy labels, our objective is to ensure that the content factors of all examples in the same underlying clean class remain unchanged as their style information changes.To achieve this, we utilize different data augmentation techniques to alter the styles while regularizing content factors based on some confident examples.